TALKING HEADS: Physical, Linguistic and Cognitive Issues in Facial Animation
نویسندگان
چکیده
Facial modeling and animation are increasingly receiving attention in the graphics and artificial intelligence (AI) research communities, both of which share the common goal of synthesizing believable, simulated agents. While computer graphics researchers have been primarily concerned with physical and anatomical aspects of facial movements, AI researchers and cognitive scientists have focused on understanding and modeling the motivation behind those movements and expressions. The combination of these two avenues of research may eventually lead to agents that can interact autonomously, with humans or with each other, bearing faces that believably model the underlying meaning of the interactions. While such synthetic speaking faces are undoubtedly useful for cognitive research, their practical applications are also vast in number, encompassing such diverse fields as medicine, education, telecommunications and the entertainment industry. Facial expressions have fascinated mankind for centuries, depicting our personalities, emotions, moods, and speech patterns. When interacting with one another, we examine faces to gather information that may not be conveyed by the spoken words alone. We are able to recognize faces from our earliest days of childhood, making them perhaps our most basic communicative tool. In fact, it is precisely this familiarity with facial expressions that makes it difficult for researchers to simulate them, for we are always acutely aware of the smallest discrepancies between real and synthetic faces. The purpose of this course is to give on overview of the current state of the art in the field of facial modeling and animation, paying particular attention to the relevant linguistic and cognitive issues in determining facial expressions. In Part I of the course, we lay the groundwork for discussing facial animation by presenting an introduction to facial anatomy and the physical aspects of facial modeling. In Part II, we describe the role of intonational patterns in speech, which partially determine the facial movements that accurately reflect the semantic content of the speech. Finally, in Part III, we describe the emotional and linguistic aspects of facial expression and describe an implemented system for automatically generating speech with context-appropriate intonation and corresponding facial animation.
منابع مشابه
Compression of MPEG-4 facial animation parameters for transmission of talking heads
The emerging MPEG-4 standard supports the transmission and composition of facial animation with natural video. The new standard will include a facial animation parameter (FAP) set that is defined based on the study of minimal facial actions and is closely related to muscle actions. The FAP set enables model-based representation of natural or synthetic talking-head sequences and allows intelligi...
متن کاملT ALKING H EADS : Physical , Linguistic and Cognitive Issues in Facial Animation Course
Facial modeling and animation are increasingly receiving attention in the graphics and artificial intelligence (AI) research communities, both of which share the common goal of synthesizing believable, simulated agents. While computer graphics researchers have been primarily concerned with physical and anatomical aspects of facial movements, AI researchers and cognitive scientists have focused ...
متن کاملTalking Head: Synthetic Video Facial Animation in MPEG-4
We present a system for facial modeling and animation that aims at the generation of photo-realistic models and performance driven animation. It is practical implementation of MPEG-4 compliant Synthetic Video Facial Animation pipeline (Simple and Calibration Profiles with some modifications), which includes: facial features recognition & tracking on real video sequence; obtaining, encoding, net...
متن کاملReal-time streaming for the animation of talking faces in multiuser environments
In order to enable face animation on the Internet using high quality synthetic speech, the Text-to-Speech (TTS) servers need to be implemented on network-based servers and shared by many users. The output of a TTS server is used to animate talking heads as defined in MPEG-4. The TTS server creates two sets of data: audio data and Phonemes with optional Facial Animation Parameters (FAP) like smi...
متن کاملGeneration of Personalized MPEG-4 compliant Talking Heads
This paper studies a new method for three-dimensional (3D) facial model adaptation and its integration into a Text-to-Speech (TTS) system. The TTS System pronounces, in real time, English or Greek speech and simultaneously animates the adapted face model, thus simulating a natural talking face. The 3D facial adaptation requires a set of two orthogonal views of the user’s face with a number of f...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1995